Cookin’: Interactive Cooking Assistant
نویسنده
چکیده
This report focuses on an interactive system that helps people during cooking by guiding them through the steps of a recipe which is obtained with the Spoonacular API. This involves both text and speech instructions that the system gives to the user, while the user can interact with it by using their voice and poses. In literature, these systems have been explored before, but only certain parts were implemented or it only gave limited feedback during cooking. This system guides a user completely through the process by reading out the equipment, ingredients and steps. Each multimedia component, speech and visuals, will be explored separately and then each component will be connected to the actual system by a user interface. This results in a system that works with textual input at the start, and voice commands and poses during the process where machine learning is used for the recognition of poses and the speech is made more efficient by using more words to recognize commands. Speech recognition is done with the Google Speech Recognition API and Kinect is used for the detection of the joint coordinates of poses. Training is done on a dataset which contains the arm coordinates and that can be extended. For classifying, the k-nearest neighbor algorithm was used since it gives the highest accuracy (99.76%) and is also relatively fast. The value for k is chosen to be 8, based on the amount of poses and a noise class. In the future, other multimedia components can be added to the infrastructure, since it is easily adjustable.
منابع مشابه
What's Cookin'? Interpreting Cooking Videos using Text, Speech and Vision
We present a novel method for aligning a sequence of instructions to a video of someone carrying out a task. In particular, we focus on the cooking domain, where the instructions correspond to the recipe. Our technique relies on an HMM to align the recipe steps to the (automatically generated) speech transcript. We then refine this alignment using a state-of-the-art visual food detector, based ...
متن کاملManagement of Distributed RFID Surfaces: A Cooking Assistant for Ambient Computing in Kitchen
Interactive tabletops are used to support interaction when designing applications. The design is generally displayed on the detection surface, but in some cases, the display uses other surfaces in a distributed way. This paper proposes some additions of an existing architecture. This new architecture aims for managing distributed surfaces with agent concepts of a Multi-Agent System. The Multi-A...
متن کاملAutomatic Generation of Interactive Cooking Video with Semantic Annotation
Videos are one of the most frequently used forms of multimedia resources. People want to interact with videos to find a specific part or to obtain relevant information. To support user interactions, current videos should be transformed to interactive videos. This paper proposes an interactive cooking video system to generate automatically interactive cooking videos. To do this, the proposed sys...
متن کاملBehavior-Sensitive User Interfaces for Smart Environments
In smart environments interactive assistants can support the user’s daily life by being ubiquitously available through any interaction device that is connected to the network. Focusing on graphical interaction, user interfaces are required to be flexible enough to be adapted to the actual context of the user. In this paper we describe an approach, which enables flexible user interface layout ad...
متن کاملIngredients and Recipe for a Robust Mobile Speech-Enabled Cooking Assistant for German
We describe the concept and implementation of Kochbot, a cooking assistant application for smartphones and tablet devices that robustly processes speech I/O and supports German recipes. Its main functions are (1) helping searching in a large recipe collection, (2) reading out loud the cooking instructions step-by-step, and (3) answering questions during cooking. Our goal was to investigate and ...
متن کامل